Parallel Successive Convex Approximation for Nonsmooth Nonconvex Optimization
نویسندگان
چکیده
Consider the problem of minimizing the sum of a smooth (possibly non-convex) and a convex (possibly nonsmooth) function involving a large number of variables. A popular approach to solve this problem is the block coordinate descent (BCD) method whereby at each iteration only one variable block is updated while the remaining variables are held fixed. With the recent advances in the developments of the multi-core parallel processing technology, it is desirable to parallelize the BCD method by allowing multiple blocks to be updated simultaneously at each iteration of the algorithm. In this work, we propose an inexact parallel BCD approach where at each iteration, a subset of the variables is updated in parallel by minimizing convex approximations of the original objective function. We investigate the convergence of this parallel BCD method for both randomized and cyclic variable selection rules. We analyze the asymptotic and non-asymptotic convergence behavior of the algorithm for both convex and non-convex objective functions. The numerical experiments suggest that for a special case of Lasso minimization problem, the cyclic block selection rule can outperform the randomized rule.
منابع مشابه
Benson's algorithm for nonconvex multiobjective problems via nonsmooth Wolfe duality
In this paper, we propose an algorithm to obtain an approximation set of the (weakly) nondominated points of nonsmooth multiobjective optimization problems with equality and inequality constraints. We use an extension of the Wolfe duality to construct the separating hyperplane in Benson's outer algorithm for multiobjective programming problems with subdifferentiable functions. We also fo...
متن کاملAsynchronous Parallel Algorithms for Nonconvex Big-Data Optimization. Part I: Model and Convergence
We propose a novel asynchronous parallel algorithmic framework for the minimization of the sum of a smooth nonconvex function and a convex nonsmooth regularizer, subject to both convex and nonconvex constraints. The proposed framework hinges on successive convex approximation techniques and a novel probabilistic model that captures key elements of modern computational architectures and asynchro...
متن کاملAsynchronous Parallel Algorithms for Nonconvex Big-Data Optimization Part II: Complexity and Numerical Results
We present complexity and numerical results for a new asynchronous parallel algorithmic method for the minimization of the sum of a smooth nonconvex function and a convex nonsmooth regularizer, subject to both convex and nonconvex constraints. The proposed method hinges on successive convex approximation techniques and a novel probabilistic model that captures key elements of modern computation...
متن کاملDistributed Methods for Constrained Nonconvex Multi-Agent Optimization-Part I: Theory
In this two-part paper, we propose a general algorithmic framework for the minimization of a nonconvex smooth function subject to nonconvex smooth constraints. The algorithm solves a sequence of (separable) strongly convex problems and mantains feasibility at each iteration. Convergence to a stationary solution of the original nonconvex optimization is established. Our framework is very general...
متن کاملNonconvex Optimization for Communication Systems
Convex optimization has provided both a powerful tool and an intriguing mentality to the analysis and design of communication systems over the last few years. A main challenge today is on nonconvex problems in these application. This paper presents an overview of some of the important nonconvex optimization problems in point-to-point and networked communication systems. Three typical applicatio...
متن کامل